Skip to content

Conversation

@huydhn
Copy link
Contributor

@huydhn huydhn commented Oct 29, 2025

Purpose

This relands #27598 exactly as it is. For the CUDA 13.0 build failure in https://buildkite.com/vllm/release/builds/9637#019a3162-f87b-4fd4-820a-5612913b590e, I have added the missing xformers wheel for cu130 at https://download.pytorch.org/whl/cu130/xformers-0.0.33%2B5d4b92a5.d20251029-cp39-abi3-linux_x86_64.whl

Test Plan

https://buildkite.com/vllm/release/builds/9672 should all be green

@simon-mo

Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request reapplies the installation of a pre-built xformers wheel, removing the previous build-from-source step in the Dockerfile. However, there appears to be a critical version mismatch in the xformers dependency specified in requirements/cuda.txt. The version in the file does not match the version mentioned in the pull request description, which was specifically added to fix a build failure. This should be corrected to ensure the build works as intended.

@mergify mergify bot added the ci/build label Oct 29, 2025
Copy link

@chatgpt-codex-connector chatgpt-codex-connector bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

💡 Codex Review

Here are some automated review suggestions for this pull request.

ℹ️ About Codex in GitHub

Codex has been enabled to automatically review pull requests in this repo. Reviews are triggered when you

  • Open a pull request for review
  • Mark a draft as ready
  • Comment "@codex review".

If Codex has suggestions, it will comment; otherwise it will react with 👍.

When you sign up for Codex through ChatGPT, Codex can also answer questions or update the PR, like "@codex address that feedback".

Copy link
Collaborator

@simon-mo simon-mo left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

See bot comments

Signed-off-by: Huy Do <huydhn@gmail.com>
@huydhn huydhn force-pushed the reland-use-prebuilt-xformers branch from 6f07867 to 145906b Compare October 29, 2025 21:00
@simon-mo simon-mo enabled auto-merge (squash) October 29, 2025 22:14
@github-actions github-actions bot added the ready ONLY add when PR is ready to merge/full CI is needed label Oct 29, 2025
Signed-off-by: Huy Do <huydhn@gmail.com>
auto-merge was automatically disabled October 30, 2025 01:25

Head branch was pushed to by a user without write access

@noooop
Copy link
Collaborator

noooop commented Oct 30, 2025

Sorry to interrupt, unblock language-models-test-extended-pooling expect this PR to fix the main failure, but it didn't. (This failure might not be related to xformers.)

@huydhn
Copy link
Contributor Author

huydhn commented Oct 30, 2025

Sorry to interrupt, unblock language-models-test-extended-pooling expect this PR to fix the main failure, but it didn't. (This failure might not be related to xformers.)

It makes sense that they are not related. This PR just swaps in a prebuilt xformers wheel instead of building it from source in exactly the same way.

@noooop
Copy link
Collaborator

noooop commented Oct 30, 2025

Sorry to interrupt, unblock language-models-test-extended-pooling expect this PR to fix the main failure, but it didn't. (This failure might not be related to xformers.)

It makes sense that they are not related. This PR just swaps in a prebuilt xformers wheel instead of building it from source in exactly the same way.

I don't know why #27659 can succeed; what is the subtle difference between it and the main requirements?

https://buildkite.com/vllm/ci/builds/36673/steps/canvas?sid=019a2cf2-4410-4d56-9eec-27dd93dfe177

@huydhn
Copy link
Contributor Author

huydhn commented Oct 30, 2025

@simon-mo I think those failures are the same nightly failures from before while https://buildkite.com/vllm/release/builds/9677 looks ok now

@simon-mo simon-mo merged commit ba33e88 into vllm-project:main Oct 30, 2025
82 of 91 checks passed
@0xC45 0xC45 mentioned this pull request Oct 31, 2025
1 task
luccafong added a commit to luccafong/vllm that referenced this pull request Nov 3, 2025
ZhengHongming888 pushed a commit to ZhengHongming888/vllm that referenced this pull request Nov 8, 2025
@ziruiliu
Copy link
Contributor

ziruiliu commented Nov 8, 2025

xformers==0.0.33+5d4b92a5.d20251029 only appears in cuda 13.0 prebuilt wheel.
For most people using cuda 12.X, the change breaks the installation

rtourgeman pushed a commit to rtourgeman/vllm that referenced this pull request Nov 10, 2025
luccafong added a commit to luccafong/vllm that referenced this pull request Nov 11, 2025
eldarkurtic pushed a commit to eldarkurtic/vllm that referenced this pull request Nov 12, 2025
…vllm-project#27768)

Signed-off-by: Huy Do <huydhn@gmail.com>
Signed-off-by: Eldar Kurtic <8884008+eldarkurtic@users.noreply.github.com>
@tzulingk
Copy link

when I tried to use deepseek-ai/DeepSeek-V2-Lite. I get the following error:

  File "/workspace/components/src/dynamo/vllm/main.py", line 115, in worker
    await init(runtime, config)
  File "/workspace/components/src/dynamo/vllm/main.py", line 413, in init
    engine_client, vllm_config, default_sampling_params = setup_vllm_engine(
                                                          ^^^^^^^^^^^^^^^^^^
  File "/workspace/components/src/dynamo/vllm/main.py", line 208, in setup_vllm_engine
    engine_args.create_model_config().get_diff_sampling_param()
    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/opt/vllm/vllm/engine/arg_utils.py", line 1186, in create_model_config
    return ModelConfig(
           ^^^^^^^^^^^^
  File "/opt/dynamo/venv/lib/python3.12/site-packages/pydantic/_internal/_dataclasses.py", line 123, in __init__
    s.__pydantic_validator__.validate_python(ArgsKwargs(args, kwargs), self_instance=s)
pydantic_core._pydantic_core.ValidationError: 1 validation error for ModelConfig
  Value error, Model architectures ['DeepseekV2ForCausalLM'] failed to be inspected. Please check the logs for more details. [type=value_error, input_value=ArgsKwargs((), {'model': ...rocessor_plugin': None}), input_type=ArgsKwargs]
    For further information visit https://errors.pydantic.dev/2.11/v/value_error

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ci/build ready ONLY add when PR is ready to merge/full CI is needed

Projects

None yet

Development

Successfully merging this pull request may close these issues.

5 participants